Neural variational sparse topic model for sparse explainable text representation

نویسندگان

چکیده

Texts are the major information carrier for internet users, from which learning latent representations has important research and practical value. Neural topic models have been proposed great performance in extracting interpretable topics of texts. However, there remain two limitations: (1) these methods generally ignore contextual texts limited feature representation ability due to shallow feed-forward network architecture, (2) Sparsity semantic space is ignored. To address issues, this paper, we propose a reinforcement neural variational sparse model (SR-NSTM) towards explainable text learning. Compared with existing models, SR-NSTM generative process probabilistic distributions parameterized networks incorporates Bi-directional LSTM embed at document level. It achieves posterior over documents words zero-mean Laplace distribution sparsemax. Moreover, supervised extension via adding max-margin regularization tackle tasks. The inference method utilized learn our efficiently. Experimental results on Web Snippets, 20Newsgroups, BBC, Biomedical datasets demonstrate that revisiting can improve performance, leading competitive coherent • We representation. extend tasks constraints. superiority perplexity, coherence classification accuracy.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A New IRIS Segmentation Method Based on Sparse Representation

Iris recognition is one of the most reliable methods for identification. In general, itconsists of image acquisition, iris segmentation, feature extraction and matching. Among them, iris segmentation has an important role on the performance of any iris recognition system. Eyes nonlinear movement, occlusion, and specular reflection are main challenges for any iris segmentation method. In thi...

متن کامل

A New IRIS Segmentation Method Based on Sparse Representation

Iris recognition is one of the most reliable methods for identification. In general, itconsists of image acquisition, iris segmentation, feature extraction and matching. Among them, iris segmentation has an important role on the performance of any iris recognition system. Eyes nonlinear movement, occlusion, and specular reflection are main challenges for any iris segmentation method. In thi...

متن کامل

Sparse Nonparametric Topic Model for Transfer Learning

Count data arises for example in bioinformatics or analysis of text documents represented as word count vectors. With several data sets available from related sources, exploiting their similarities by transfer learning can improve models compared to modeling sources independently. We introduce a Bayesian generative transfer learning model which represents similarity across document collections ...

متن کامل

Image Classification via Sparse Representation and Subspace Alignment

Image representation is a crucial problem in image processing where there exist many low-level representations of image, i.e., SIFT, HOG and so on. But there is a missing link across low-level and high-level semantic representations. In fact, traditional machine learning approaches, e.g., non-negative matrix factorization, sparse representation and principle component analysis are employed to d...

متن کامل

Sparse Supervised Topic Model: Midterm Report

In this paper we propose the sparse supervised topic model (SSTM), a graphical model that learns topic structures of a given document collection and also a sparse linear prediction model for response vairables associated with documents. Our model jointly learns the topics and the classifier and encourages a sparse classifier by concentrating all the relevant information for prediction into a sm...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Information Processing and Management

سال: 2021

ISSN: ['0306-4573', '1873-5371']

DOI: https://doi.org/10.1016/j.ipm.2021.102614